Viscosity Solutions of the Bellman Equation for Infinite Horizon Optimal Control Problems with Negative Instantaneous Costs
نویسنده
چکیده
In a series of papers, we characterized the value function in optimal control as the unique viscosity solution of the corresponding Bellman equation that satisfies appropriate side conditions. The novelty of our results was that they applied to exit time problems with general nonnegative instantaneous costs, including cases where the instantaneous cost is not uniformly bounded below by positive constants. This note will extend these results to control problems whose instantaneous costs are allowed to take both positive and negative values, including undiscounted examples. We apply our results to the generalized Zubov equation, which corresponds to the Bellman equation for a negative instantaneous cost. The unique solutions of the Zubov equations are maximum cost Lyapunov functions for perturbed asymptotically stable systems. We study the regularity of these Lyapunov functions, and we further extend Zubov’s method for representing domains of attractions as sublevel sets of Lyapunov functions. We also illustrate some special properties of maximum cost Lyapunov functions that can occur when the instantaneous cost for the Lyapunov function is degenerate.
منابع مشابه
Viscosity Solutions of the Bellman Equation for Infinite Horizon Optimal Control Problems with Negative Instantaneous Costs (I)
In a series of papers, we characterized the value function in optimal control as the unique viscosity solution of the corresponding Bellman equation that satisfies appropriate side conditions. The novelty of our results was that they applied to exit time problems with general nonnegative instantaneous costs, including cases where the instantaneous cost is not uniformly bounded below by positive...
متن کاملFurther Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs
We study the Bellman equation for undiscounted exit time optimal control problems for nonlinear systems and nonnegative instantaneous costs using the dynamic programming approach. Using viscosity solution theory, we prove a uniqueness theorem that characterizes the value functions for these problems as the unique viscosity solutions of the corresponding Bellman equations that satisfy appropriat...
متن کاملOn the value function for optimal control problems with infinite horizon
In this paper we consider nonautonomous optimal control problems of infinite horizon type, whose control actions are given by L-functions. We verify that the value function is locally Lipschitz. The equivalence between dynamic programming inequalities and HJB inequalities for proximal sub (super) gradients is proven. Using this result we show that the value function is a Dini solution of the HJ...
متن کاملA Bellman Approach for Regional Optimal Control Problems in RN
This article is a continuation of a previous work where we studied infinite horizon control problems for which the dynamic, running cost and control space may be different in two halfspaces of some euclidian space R . In this article we extend our results in several directions: (i) to more general domains; (ii) by considering finite horizon control problems; (iii) by weaken the controlability a...
متن کاملStochastic Optimal Control Problems with a Bounded Memory∗
This paper treats a finite time horizon optimal control problem in which the controlled state dynamics is governed by a general system of stochastic functional differential equations with a bounded memory. An infinite-dimensional HJB equation is derived using a Bellman-type dynamic programming principle. It is shown that the value function is the unique viscosity solution of the HJB equation. I...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2002